e support vector machine algorithm
vector machine (SVM) was proposed by Vladimir Vapnik [Cortes
nik, 1995]. The basic principle of SVM is to use a kernel function
raw data space to a kernel space, where classification analysis or
n analysis can be implemented [Aizerman, et al., 1964; Cortes
nik, 1995; Bishop, 2006]. It is also hypothesised that such a
space is linear.
FNN, all radial basis functions will have their weights estimated
ero values. In reality, some radial basis functions may be
t. In addition, the maximised likelihood approach [Duda, et al.,
hop, 2006; Rossi, 2018] used to estimate model parameters using
squared errors approach may not be able to generate an optimal
searching for the best hyperplane for the discrimination between
es of data points, SVM finds a set of data points which are most
o classify. These data points are referred to as the support vectors.
the closest to the hyperplane and are located on the boundaries
ssification margin between two classes.
dea proposed in SVM is to optimise the classification margin in a
space. Figure 3.29 shows this principle. In Figure 3.29(a), a wider
tion margin will allow some data points which can be easily
to fall in. In Figure 3.29(b), no such data point is found within
ification margin. SVM will optimise the classification margin
he search of the support vectors. No data points other than support
will be used as kernels, hence the model is more parsimonious than
N model.
M classifier is more robust to data noise or outliers due to the
ion of a classification margin. The generalisation capability is
boosted. A trained SVM classifier is a linear combination of the
es between an input and support vectors. The similarity between
vector and a support vector can be quantified by a kernel function,
defined as below, where ܠ is the nth support vector,